Build an open, managed and high-performance Iceberg lakehouse to enable advanced analytics and data science, with automated data management and built-in governance.
Features
BigLake Metastore is a serverless metastore for all your Iceberg tables. Engines like Apache Spark, BigQuery and third party platforms can use it to create and manage tables, giving you a consistent view of your data and unified access controls. BigLake Metastore now supports the Apache Iceberg Rest Catalog (Preview) for easy integration with OSS and third party engines. Iceberg tables are now data accessible in AlloyDB (preview), to gain interoperability between transactional and analytical platforms.
BigLake extends Google Cloud Storage management capabilities, enabling you to use storage auto-class for efficient cold data tiering and apply customer-managed encryption keys (CMEK) to your storage buckets. BigLake metastore natively integrates into Dataplex Universal Catalog, ensuring that governance policies defined centrally are consistently enforced across multiple engines while enabling semantic search, data lineage, profiling, and quality checks.
BigLake tables for Apache Iceberg offer an enterprise-ready, fully managed Iceberg experience when used with BigQuery. By storing Apache Iceberg data in your own Google Cloud Storage buckets and leveraging BigQuery's highly scalable, real time metadata management capabilities, you get the best of both worlds - the openness and data ownership associated with GCS as well as access to BigQuery's fully managed capabilities with Iceberg data for streaming, advance analytics and AI use cases.
How It Works
Common Uses
Understanding the Google Cloud components of an open data lakehouse
To build an Iceberg lakehouse with BigLake, start by storing your data in Cloud Storage. Then, define this data using BigLake tables for Apache Iceberg. The BigLake metastore serves as your centralized, serverless catalog for these Iceberg tables, eliminating the need to manage complex infrastructure. This setup allows any Iceberg-compatible engine to consistently access and manage your data, creating a unified, open, and scalable lakehouse environment with ease.
Understanding the Google Cloud components of an open data lakehouse
To build an Iceberg lakehouse with BigLake, start by storing your data in Cloud Storage. Then, define this data using BigLake tables for Apache Iceberg. The BigLake metastore serves as your centralized, serverless catalog for these Iceberg tables, eliminating the need to manage complex infrastructure. This setup allows any Iceberg-compatible engine to consistently access and manage your data, creating a unified, open, and scalable lakehouse environment with ease.
Provide real-time insights and predictions for financial services
You can use Apache Iceberg for evolving data lake datasets like transactions or market feeds. BigLake enables BigQuery to then query Iceberg tables alongside native storage without data movement. You can ingest real-time streams into BigQuery, combining with historical Iceberg data via BigLake for immediate, comprehensive analysis. BigQuery ML then generates real-time insights like market volatility and fraud detection as well as predictive models like credit risk and customer behavior.
Provide real-time insights and predictions for financial services
You can use Apache Iceberg for evolving data lake datasets like transactions or market feeds. BigLake enables BigQuery to then query Iceberg tables alongside native storage without data movement. You can ingest real-time streams into BigQuery, combining with historical Iceberg data via BigLake for immediate, comprehensive analysis. BigQuery ML then generates real-time insights like market volatility and fraud detection as well as predictive models like credit risk and customer behavior.
BigLake provides secure, consistent access to a single copy of data in Cloud Storage. Dataplex Universal Catalog then automatically catalogs this data so all data users and engines can access. This ensures consistent data definitions, easy discovery, and unified governance, eliminating silos and fostering collaboration on one source of truth.
BigLake provides secure, consistent access to a single copy of data in Cloud Storage. Dataplex Universal Catalog then automatically catalogs this data so all data users and engines can access. This ensures consistent data definitions, easy discovery, and unified governance, eliminating silos and fostering collaboration on one source of truth.
Pricing
How BigLake pricing works | BigLake pricing is based on table management, metadata storage and metadata access | |
---|---|---|
Services and usage | Description | Price (USD) |
BigLake table management | BigLake table management compute resources used for automatic table storage optimization. | Starting at $0.12 per DCU-Hour |
BigLake metadata storage | BigLake Metastore charges for metadata stored. Free tier includes 1 GiB of metadata storage per month included. | Starting at $0.04 per GiB per month |
BigLake metadata access | Class A Operations: BigLake metadata access charges for writes, updates, list, create, and config operations with a free tier of 5,000 operations per month included. | Starting at $6.00 per million operations |
Class B Operations: BigLake metadata access charges for reads, get, and delete operations with a free tier of 50,000 operations per month included. | Starting at $0.90 per million operations |
How BigLake pricing works
BigLake pricing is based on table management, metadata storage and metadata access
BigLake table management
BigLake table management compute resources used for automatic table storage optimization.
Starting at
$0.12
per DCU-Hour
BigLake metadata storage
BigLake Metastore charges for metadata stored. Free tier includes 1 GiB of metadata storage per month included.
Starting at
$0.04
per GiB per month
BigLake metadata access
Class A Operations: BigLake metadata access charges for writes, updates, list, create, and config operations with a free tier of 5,000 operations per month included.
Starting at
$6.00
per million operations
Class B Operations: BigLake metadata access charges for reads, get, and delete operations with a free tier of 50,000 operations per month included.
Starting at
$0.90
per million operations